Asymptotic Generalization Bound of Fisher's Linear Discriminant Analysis

نویسندگان

  • Wei Bian
  • Dacheng Tao
چکیده

Fisher's linear discriminant analysis (FLDA) is an important dimension reduction method in statistical pattern recognition. It has been shown that FLDA is asymptotically Bayes optimal under the homoscedastic Gaussian assumption. However, this classical result has the following two major limitations: 1) it holds only for a fixed dimensionality D, and thus does not apply when D and the training sample size N are proportionally large; 2) it does not provide a quantitative description on how the generalization ability of FLDA is affected by D and N. In this paper, we present an asymptotic generalization analysis of FLDA based on random matrix theory, in a setting where both D and N increase and D/N → γ ∈ [0,1). The obtained lower bound of the generalization discrimination power overcomes both limitations of the classical result, i.e., it is applicable when D and N are proportionally large and provides a quantitative description of the generalization ability of FLDA in terms of the ratio γ = D/N and the population discrimination power. Besides, the discrimination power bound also leads to an upper bound on the generalization error of binary-classification with FLDA.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Non-asymptotic Analysis of Compressive Fisher Discriminants in terms of the Effective Dimension

We provide a non-asymptotic analysis of the generalisation error of compressive Fisher linear discriminant (FLD) classification that is dimension free under mild assumptions. Our analysis includes the effects that random projection has on classification performance under covariance model misspecification, as well as various good and bad effects of random projections that contribute to the overa...

متن کامل

LS Bound based gene selection for DNA microarray data

MOTIVATION One problem with discriminant analysis of DNA microarray data is that each sample is represented by quite a large number of genes, and many of them are irrelevant, insignificant or redundant to the discriminant problem at hand. Methods for selecting important genes are, therefore, of much significance in microarray data analysis. In the present study, a new criterion, called LS Bound...

متن کامل

Error bounds for Kernel Fisher Linear Discriminant in Gaussian Hilbert space

We give a non-trivial, non-asymptotic upper bound on the classification error of the popular Kernel Fisher Linear Discriminant classifier under the assumption that the kernelinduced space is a Gaussian Hilbert space.

متن کامل

An algorithm to compute relative cubic fields

Let K be an imaginary quadratic number field with class number 1. We describe a new, essentially linear-time algorithm, to list all isomorphism classes of cubic extensions L/K up to a bound X on the norm of the relative discriminant ideal. The main tools are Taniguchi’s [18] generalization of Davenport-Heilbronn parametrisation of cubic extensions, and reduction theory for binary cubic forms ov...

متن کامل

Vision and Pattern Recognition , San Francisco , 1996 . Pattern Rejection

The e ciency of pattern recognition is particularly crucial in two scenarios; whenever there are a large number of classes to discriminate, and, whenever recognition must be performed a large number of times. We propose a single technique, namely, pattern rejection, that greatly enhances e ciency in both cases. A rejector is a generalization of a classi er, that quickly eliminates a large fract...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • IEEE transactions on pattern analysis and machine intelligence

دوره 36 12  شماره 

صفحات  -

تاریخ انتشار 2014